Crowdsourcing for Social Multimedia at MediaEval 2013: Challenges, Data set, and Evaluation
نویسندگان
چکیده
This paper provides an overview of the Crowdsourcing for Multimedia Task at MediaEval 2013 multimedia benchmarking initiative. The main goal of this task is to assess the potential of hybrid human/conventional computation techniques to generate accurate labels for social multimedia content. The task data are fashion-related images, collected from the Web-based photo sharing platform Flickr. Each image is accompanied by a) its metadata (e.g., title, description, and tags), and b) a set of ‘basic human labels’ collected from human annotators using a microtask with a basic quality control mechanism that is run on the Amazon Mechanical Turk crowdsourcing platform. The labels reflect whether or not the image depicts fashion, and whether or not the image matches its ‘category’ (i.e., the fashion-related query that returned the image from Flickr). The ‘basic human labels’ were collected such that their noise levels would be characteristic of data gathered from crowdsourcing workers without using highly sophisticated quality control. The task asks participants to predict high-quality labels, either by aggregating the ‘basic human labels’ or by combining them with the context (i.e., the metadata) and/or the content (i.e., visual features) of the image.
منابع مشابه
Retrieving Diverse Social Images at MediaEval 2013: Objectives, Dataset and Evaluation
This paper provides an overview of the Retrieving Diverse Social Images task that is organized as part of the MediaEval 2013 Benchmarking Initiative for Multimedia Evaluation. The task addresses the problem of result diversification in the context of social photo retrieval. We present the task challenges, the proposed data set and ground truth, the required participant runs and the evaluation m...
متن کاملL3S at MediaEval 2013 Crowdsourcing for Social Multimedia Task
In this paper we present results of our initial research on aggregating noisy crowdsourced labels, by using a modified version of the EM algorithm introduced in [1]. We propose different methods of estimating the worker confidence, a measure that indicates how well the worker is performing the task, and of integrating it in the computation of the aggregated label. Furthermore, we introduce a no...
متن کاملRetrieving Diverse Social Images at MediaEval 2015: Challenge, Dataset and Evaluation
This paper provides an overview of the Retrieving Diverse Social Images task that is organized as part of the MediaEval 2015 Benchmarking Initiative for Multimedia Evaluation. The task addresses the problem of result diversification and user annotation credibility estimation in the context of social photo retrieval. We present the task challenges, the proposed data set and ground truth, the req...
متن کاملRetrieving Diverse Social Images at MediaEval 2014: Challenge, Dataset and Evaluation
This paper provides an overview of the Retrieving Diverse Social Images task that is organized as part of the MediaEval 2014 Benchmarking Initiative for Multimedia Evaluation. The task addresses the problem of result diversification in the context of social photo retrieval. We present the task challenges, the proposed data set and ground truth, the required participant runs and the evaluation m...
متن کاملSQUARE: Benchmarking Crowd Consensus at MediaEval
We extend the square benchmark for statistical consensus methods to include additional evaluation on two datasets from the MediaEval 2013 Crowdsourcing in Multimedia shared task. In addition to reporting shared task results, we also analyze qualitatively and quantitatively performance of consensus algorithms under varying supervision. 1. ALGORITHMS We extend square [5], a benchmark for evaluati...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2013